Goto

Collaborating Authors

 attention augmented agent



Towards Interpretable Reinforcement Learning Using Attention Augmented Agents

Neural Information Processing Systems

Inspired by recent work in attention models for image captioning and question answering, we present a soft attention model for the reinforcement learning domain. This model bottlenecks the view of an agent by a soft, top-down attention mechanism, forcing the agent to focus on task-relevant information by sequentially querying its view of the environment. The output of the attention mechanism allows direct observation of the information used by the agent to select its actions, enabling easier interpretation of this model than of traditional models. We analyze the different strategies the agents learn and show that a handful of strategies arise repeatedly across different games. We also show that the model learns to query separately about space and content ( what''). We demonstrate that an agent using this mechanism can achieve performance competitive with state-of-the-art models on ATARI tasks while still being interpretable.


Reviews: Towards Interpretable Reinforcement Learning Using Attention Augmented Agents

Neural Information Processing Systems

The paper is well-written and clear; the architecture is described in detail through a diagram (Figure 1 on page 2), with the math in section 2 expanding on the key components of the attention mechanism. High-level details for the RL training setup, implemented baselines, and condensed results are provided in the body of the paper. Detailed learning curves for each of the compared approaches are presented in the appendix (which is appropriate, given that the task-specific learning performance is secondary to the analysis of the attention mechanism). The analysis section is thorough, and I specifically appreciated the section at the end comparing the learned attention mechanism to prior work on saliency maps. Model/Architecture Notes: While the proposed model is a straightforward extension of query-key-value attention to tasks in RL, there are two interesting architectural features: First, "queries" for their attention mechanism can be decomposed into features that act on content (which the paper refers to as the "what"), and features that act on spatial location (which the paper refers to as the "where").


Reviews: Towards Interpretable Reinforcement Learning Using Attention Augmented Agents

Neural Information Processing Systems

The reviewers appreciated the analysis of the learned attention from the model, and recommend accepting. I would ask that the authors add some of the experiments suggested by R4 in their "key questions" section. If there is not enough space, please add these to the supplemental.


Towards Interpretable Reinforcement Learning Using Attention Augmented Agents

Neural Information Processing Systems

Inspired by recent work in attention models for image captioning and question answering, we present a soft attention model for the reinforcement learning domain. This model bottlenecks the view of an agent by a soft, top-down attention mechanism, forcing the agent to focus on task-relevant information by sequentially querying its view of the environment. The output of the attention mechanism allows direct observation of the information used by the agent to select its actions, enabling easier interpretation of this model than of traditional models. We analyze the different strategies the agents learn and show that a handful of strategies arise repeatedly across different games. We also show that the model learns to query separately about space and content (where'' vs.what'').


Towards Interpretable Reinforcement Learning Using Attention Augmented Agents

Mott, Alexander, Zoran, Daniel, Chrzanowski, Mike, Wierstra, Daan, Rezende, Danilo Jimenez

Neural Information Processing Systems

Inspired by recent work in attention models for image captioning and question answering, we present a soft attention model for the reinforcement learning domain. This model bottlenecks the view of an agent by a soft, top-down attention mechanism, forcing the agent to focus on task-relevant information by sequentially querying its view of the environment. The output of the attention mechanism allows direct observation of the information used by the agent to select its actions, enabling easier interpretation of this model than of traditional models. We analyze the different strategies the agents learn and show that a handful of strategies arise repeatedly across different games. We also show that the model learns to query separately about space and content ( where'' vs. what'').